AdaMoW: Multimodal Sentiment Analysis Based on Adaptive Modality-Specific Weight Fusion Network
نویسندگان
چکیده
Multimodal sentiment analysis (MSA) is a crucial task in the field of natural language processing (NLP), with wide range applications. This paper proposes an adaptive modality-specific weight fusion network (AdaMoW) to address issues process multimodal data fusion. Specifically, we use different calculation methods at various stages model. In model training stage, diverse weights are assigned modalities by calculating correlation between single-modal prediction value and real labels, weight-mapping designed learn this “data-weight” mapping relationship. testing verification phase model, trained used obtain modalities. addition, order optimize data, generator, which reversely generates unimodal feature vector through vector, compares it original extraction obtained after extraction. The modal vectors compared optimized, so that results can maintain uniqueness modality while obtaining interaction information. AdaMoW verified on two benchmark MSA datasets CMU-MOSI CMU-MOSEI. experimental show effectiveness surpasses previous baseline achieves state-of-the-art results.
منابع مشابه
Tensor Fusion Network for Multimodal Sentiment Analysis
Multimodal sentiment analysis is an increasingly popular research area, which extends the conventional language-based definition of sentiment analysis to a multimodal setup where other relevant modalities accompany language. In this paper, we pose the problem of multimodal sentiment analysis as modeling intra-modality and inter-modality dynamics. We introduce a novel model, termed Tensor Fusion...
متن کاملMultimodal medical image fusion based on Yager’s intuitionistic fuzzy sets
The objective of image fusion for medical images is to combine multiple images obtained from various sources into a single image suitable for better diagnosis. Most of the state-of-the-art image fusing technique is based on nonfuzzy sets, and the fused image so obtained lags with complementary information. Intuitionistic fuzzy sets (IFS) are determined to be more suitable for civilian, and medi...
متن کاملAdaptive Multimodal Fusion
Multimodal interfaces offer its users the possibility of interacting with computers, in a transparent, natural way, by means of various modalities. Fusion engines are key components in multimodal systems, responsible for combining information from different sources and extract a semantic meaning from them. This fusion process allows many modalities to be effectively used at once and therefore a...
متن کاملMultimodal Sentiment Analysis
With more than 10,000 new videos posted online every day on social websites such as YouTube and Facebook, the internet is becoming an almost infinite source of information. One important challenge for the coming decade is to be able to harvest relevant information from this constant flow of multimodal data. In this talk, I will introduce the task of multimodal sentiment analysis, and present a ...
متن کاملBenchmarking Multimodal Sentiment Analysis
We propose a framework for multimodal sentiment analysis and emotion recognition using convolutional neural network-based feature extraction from text and visual modalities. We obtain a performance improvement of 10% over the state of the art by combining visual, text and audio features. We also discuss some major issues frequently ignored in multimodal sentiment analysis research: the role of ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Access
سال: 2023
ISSN: ['2169-3536']
DOI: https://doi.org/10.1109/access.2023.3276932